# 4-bit efficient quantization
GLM 4 9B 0414 4bit DWQ
Apache-2.0
A high-performance 4-bit DWQ quantized version of GLM-4-9B, optimized for Apple chips and supporting 128K long context.
Large Language Model
G
Narutoouz
194
1
Qwen3 8b 192k Context 6X Josiefied Uncensored MLX AWQ 4bit
Apache-2.0
The 4-bit AWQ quantized version of Qwen3-8B, optimized for the MLX library, supports 192k token long context processing, suitable for edge device deployment.
Large Language Model
Q
Goraint
204
1
Featured Recommended AI Models